A large Czech language model based on continuous pre-training of the English MPT7b model, trained on 272 billion tokens of Czech corpus using a Czech tokenizer for pre-training on approximately 67 billion tokens of Czech large-scale corpus
Large Language Model Other